80 research outputs found

    An informatics system for exploring eye movements in reading

    Get PDF
    Eye tracking techniques have been widely used in many research areas including cognitive science, psychology, human-computer interaction, marketing research, medical research etc. Many computer programs have emerged to help these researchers to design experiments, present visual stimuli and process the large quantity of numerical data produced by the eye tracker. However, most applications, especially commercial products, are designed for a particular tracking device and tend to be general purpose. Few of them are designed specifically for reading research. This can be inconvenient when dealing with complex experimental design, multi-source data collection, and text based data analysis, including almost every aspect of a reading study lifecycle. A flexible and powerful system that manages the lifecycle of different reading studies is required to fulfill these demands. Therefore, we created an informatics system with two major software suites: Experiment Executor and EyeMap. It is a system designed specifically for reading research. Experiment Executor helps reading researchers build complex experimental environments, which can rapidly present display changes and support the co-registration of eye tracking information with other data collection devices such as EEG (electroencephalography) amplifiers. The EyeMap component helps researchers visualize and analysis a wide range of writing systems including spaced and unspaced scripts, which can be presented in proportional or non-proportional font types. The aim of the system is to accelerate the life cycle of a reading experiment from design through analysis. Several experiments were conducted on this system. These experiments confirmed the effectiveness and the capability of the system with several new reading research findings from the visual information processing stages of reading

    SGL-PT: A Strong Graph Learner with Graph Prompt Tuning

    Full text link
    Recently, much exertion has been paid to design graph self-supervised methods to obtain generalized pre-trained models, and adapt pre-trained models onto downstream tasks through fine-tuning. However, there exists an inherent gap between pretext and downstream graph tasks, which insufficiently exerts the ability of pre-trained models and even leads to negative transfer. Meanwhile, prompt tuning has seen emerging success in natural language processing by aligning pre-training and fine-tuning with consistent training objectives. In this paper, we identify the challenges for graph prompt tuning: The first is the lack of a strong and universal pre-training task across sundry pre-training methods in graph domain. The second challenge lies in the difficulty of designing a consistent training objective for both pre-training and downstream tasks. To overcome above obstacles, we propose a novel framework named SGL-PT which follows the learning strategy ``Pre-train, Prompt, and Predict''. Specifically, we raise a strong and universal pre-training task coined as SGL that acquires the complementary merits of generative and contrastive self-supervised graph learning. And aiming for graph classification task, we unify pre-training and fine-tuning by designing a novel verbalizer-free prompting function, which reformulates the downstream task in a similar format as pretext task. Empirical results show that our method surpasses other baselines under unsupervised setting, and our prompt tuning method can greatly facilitate models on biological datasets over fine-tuning methods

    MARIO: Model Agnostic Recipe for Improving OOD Generalization of Graph Contrastive Learning

    Full text link
    In this work, we investigate the problem of out-of-distribution (OOD) generalization for unsupervised learning methods on graph data. This scenario is particularly challenging because graph neural networks (GNNs) have been shown to be sensitive to distributional shifts, even when labels are available. To address this challenge, we propose a \underline{M}odel-\underline{A}gnostic \underline{R}ecipe for \underline{I}mproving \underline{O}OD generalizability of unsupervised graph contrastive learning methods, which we refer to as MARIO. MARIO introduces two principles aimed at developing distributional-shift-robust graph contrastive methods to overcome the limitations of existing frameworks: (i) Information Bottleneck (IB) principle for achieving generalizable representations and (ii) Invariant principle that incorporates adversarial data augmentation to obtain invariant representations. To the best of our knowledge, this is the first work that investigates the OOD generalization problem of graph contrastive learning, with a specific focus on node-level tasks. Through extensive experiments, we demonstrate that our method achieves state-of-the-art performance on the OOD test set, while maintaining comparable performance on the in-distribution test set when compared to existing approaches. The source code for our method can be found at: https://github.com/ZhuYun97/MARIOComment: 20 pages, 15 figure

    Study on Helicopter Antitorque Device Based on Cross-Flow Fan Technology

    Get PDF
    In order to improve low-altitude flight security of single-rotor helicopter, an experimental model of a helicopter antitorque device is developed for wind tunnel test. The model is based on the flow control technology of the cross-flow fan (CFF). Wind tunnel tests show that the model can produce side force. It is concluded that the influence of the CFF rotating speed, the rotor collective pitch, and the forward flight speed on the side force of the model is great. At the same time, the numerical simulation calculation method of the model has been established. Good agreement between experimental and numerical side force and power shows that results of numerical solution are reliable. Therefore, the results in actual helicopter obtained from Computational Fluid Dynamics (CFD) solution are acceptable. This proves that this antitorque device can be used for a helicopter

    Cross-relation Cross-bag Attention for Distantly-supervised Relation Extraction

    Full text link
    Distant supervision leverages knowledge bases to automatically label instances, thus allowing us to train relation extractor without human annotations. However, the generated training data typically contain massive noise, and may result in poor performances with the vanilla supervised learning. In this paper, we propose to conduct multi-instance learning with a novel Cross-relation Cross-bag Selective Attention (C2^2SA), which leads to noise-robust training for distant supervised relation extractor. Specifically, we employ the sentence-level selective attention to reduce the effect of noisy or mismatched sentences, while the correlation among relations were captured to improve the quality of attention weights. Moreover, instead of treating all entity-pairs equally, we try to pay more attention to entity-pairs with a higher quality. Similarly, we adopt the selective attention mechanism to achieve this goal. Experiments with two types of relation extractor demonstrate the superiority of the proposed approach over the state-of-the-art, while further ablation studies verify our intuitions and demonstrate the effectiveness of our proposed two techniques.Comment: AAAI 201

    Dancing Avatar: Pose and Text-Guided Human Motion Videos Synthesis with Image Diffusion Model

    Full text link
    The rising demand for creating lifelike avatars in the digital realm has led to an increased need for generating high-quality human videos guided by textual descriptions and poses. We propose Dancing Avatar, designed to fabricate human motion videos driven by poses and textual cues. Our approach employs a pretrained T2I diffusion model to generate each video frame in an autoregressive fashion. The crux of innovation lies in our adept utilization of the T2I diffusion model for producing video frames successively while preserving contextual relevance. We surmount the hurdles posed by maintaining human character and clothing consistency across varying poses, along with upholding the background's continuity amidst diverse human movements. To ensure consistent human appearances across the entire video, we devise an intra-frame alignment module. This module assimilates text-guided synthesized human character knowledge into the pretrained T2I diffusion model, synergizing insights from ChatGPT. For preserving background continuity, we put forth a background alignment pipeline, amalgamating insights from segment anything and image inpainting techniques. Furthermore, we propose an inter-frame alignment module that draws inspiration from an auto-regressive pipeline to augment temporal consistency between adjacent frames, where the preceding frame guides the synthesis process of the current frame. Comparisons with state-of-the-art methods demonstrate that Dancing Avatar exhibits the capacity to generate human videos with markedly superior quality, both in terms of human and background fidelity, as well as temporal coherence compared to existing state-of-the-art approaches.Comment: 11 pages, 3 figure

    Degeneration-Tuning: Using Scrambled Grid shield Unwanted Concepts from Stable Diffusion

    Full text link
    Owing to the unrestricted nature of the content in the training data, large text-to-image diffusion models, such as Stable Diffusion (SD), are capable of generating images with potentially copyrighted or dangerous content based on corresponding textual concepts information. This includes specific intellectual property (IP), human faces, and various artistic styles. However, Negative Prompt, a widely used method for content removal, frequently fails to conceal this content due to inherent limitations in its inference logic. In this work, we propose a novel strategy named \textbf{Degeneration-Tuning (DT)} to shield contents of unwanted concepts from SD weights. By utilizing Scrambled Grid to reconstruct the correlation between undesired concepts and their corresponding image domain, we guide SD to generate meaningless content when such textual concepts are provided as input. As this adaptation occurs at the level of the model's weights, the SD, after DT, can be grafted onto other conditional diffusion frameworks like ControlNet to shield unwanted concepts. In addition to qualitatively showcasing the effectiveness of our DT method in protecting various types of concepts, a quantitative comparison of the SD before and after DT indicates that the DT method does not significantly impact the generative quality of other contents. The FID and IS scores of the model on COCO-30K exhibit only minor changes after DT, shifting from 12.61 and 39.20 to 13.04 and 38.25, respectively, which clearly outperforms the previous methods
    • …
    corecore